Goto

Collaborating Authors

 anshumali shrivastava



a1e865a9b1065392ed6035d8ccd072d9-Paper.pdf

Neural Information Processing Systems

Unfortunately,the per-iteration cost of maintaining this adaptivedistribution for gradient estimation is more than calculating the full gradient itself, which we call the chicken-and-the-egg loop. As a result, the false impression of faster convergence in iterations, inreality,leads to slower convergence in time.


c164bbc9d6c72a52c599bbb43d8db8e1-Paper.pdf

Neural Information Processing Systems

Deep neural networks have achieved impressive performance in many areas. Designing a fast and provable method for training neural networks is a fundamental question in machine learning. The classical training method requires paying Ω(mnd) cost for both forward computation and backward computation, where m is the width of the neural network, and we are given n training points in d-dimensional space.






2fc6b8a3fc23108f184daa4759024c25-Paper-Conference.pdf

Neural Information Processing Systems

IntheDistanceOracle problem,thegoalistopreprocess nvectorsx1,x2,...,xn in a d-dimensional metric space(Xd, l) into a cheap data structure, so that given a query vectorq Xd and a subsetS [n] of the input data points, all distances q xi l forxi S canbequicklyapproximated(fasterthanthetrivial d|S|querytime).